Certified Generative AI Engineer Associate v1.0

Page:    1 / 6   
Exam contains 80 questions

After changing the response generating LLM in a RAG pipeline from GPT-4 to a model with a shorter context length that the company self-hosts, the Generative AI Engineer is getting the following error:

What TWO solutions should the Generative AI Engineer implement without changing the response generating model? (Choose two.)

  • A. Use a smaller embedding model to generate embeddings
  • B. Reduce the maximum output tokens of the new model
  • C. Decrease the chunk size of embedded documents
  • D. Reduce the number of records retrieved from the vector database
  • E. Retrain the response generating model using ALiBi


Answer : CD

A Generative Al Engineer is building a system which will answer questions on latest stock news articles.
Which will NOT help with ensuring the outputs are relevant to financial news?

  • A. Implement a comprehensive guardrail framework that includes policies for content filters tailored to the finance sector.
  • B. Increase the compute to improve processing speed of questions to allow greater relevancy analysis
  • C. Implement a profanity filter to screen out offensive language.
  • D. Incorporate manual reviews to correct any problematic outputs prior to sending to the users


Answer : B

A Generative Al Engineer is building a RAG application that answers questions about internal documents for the company SnoPen AI.
The source documents may contain a significant amount of irrelevant content, such as advertisements, sports news, or entertainment news, or content about other companies.
Which approach is advisable when building a RAG application to achieve this goal of filtering irrelevant information?

  • A. Keep all articles because the RAG application needs to understand non-company content to avoid answering questions about them.
  • B. Include in the system prompt that any information it sees will be about SnoPenAI, even if no data filtering is performed.
  • C. Include in the system prompt that the application is not supposed to answer any questions unrelated to SnoPen AI.
  • D. Consolidate all SnoPen AI related documents into a single chunk in the vector database.


Answer : C

A Generative Al Engineer has successfully ingested unstructured documents and chunked them by document sections. They would like to store the chunks in a Vector Search index. The current format of the dataframe has two columns: (i) original document file name (ii) an array of text chunks for each document.
What is the most performant way to store this dataframe?

  • A. Split the data into train and test set, create a unique identifier for each document, then save to a Delta table
  • B. Flatten the dataframe to one chunk per row, create a unique identifier for each row, and save to a Delta table
  • C. First create a unique identifier for each document, then save to a Delta table
  • D. Store each chunk as an independent JSON file in Unity Catalog Volume. For each JSON file, the key is the document section name and the value is the array of text chunks for that section


Answer : B

A Generative AI Engineer has created a RAG application which can help employees retrieve answers from an internal knowledge base, such as Confluence pages or Google Drive. The prototype application is now working with some positive feedback from internal company testers. Now the Generative Al Engineer wants to formally evaluate the system’s performance and understand where to focus their efforts to further improve the system.
How should the Generative AI Engineer evaluate the system?

  • A. Use cosine similarity score to comprehensively evaluate the quality of the final generated answers.
  • B. Curate a dataset that can test the retrieval and generation components of the system separately. Use MLflow’s built in evaluation metrics to perform the evaluation on the retrieval and generation components.
  • C. Benchmark multiple LLMs with the same data and pick the best LLM for the job.
  • D. Use an LLM-as-a-judge to evaluate the quality of the final answers generated.


Answer : B

A Generative Al Engineer has already trained an LLM on Databricks and it is now ready to be deployed.
Which of the following steps correctly outlines the easiest process for deploying a model on Databricks?

  • A. Log the model as a pickle object, upload the object to Unity Catalog Volume, register it to Unity Catalog using MLflow, and start a serving endpoint
  • B. Log the model using MLflow during training, directly register the model to Unity Catalog using the MLflow API, and start a serving endpoint
  • C. Save the model along with its dependencies in a local directory, build the Docker image, and run the Docker container
  • D. Wrap the LLM’s prediction function into a Flask application and serve using Gunicorn


Answer : B

A Generative AI Engineer developed an LLM application using the provisioned throughput Foundation Model API. Now that the application is ready to be deployed, they realize their volume of requests are not sufficiently high enough to create their own provisioned throughput endpoint. They want to choose a strategy that ensures the best cost-effectiveness for their application.
What strategy should the Generative AI Engineer use?

  • A. Switch to using External Models instead
  • B. Deploy the model using pay-per-token throughput as it comes with cost guarantees
  • C. Change to a model with a fewer number of parameters in order to reduce hardware constraint issues
  • D. Throttle the incoming batch of requests manually to avoid rate limiting issues


Answer : B

A Generative AI Engineer is building an LLM to generate article summaries in the form of a type of poem, such as a haiku, given the article content. However, the initial output from the LLM does not match the desired tone or style.
Which approach will NOT improve the LLM’s response to achieve the desired response?

  • A. Provide the LLM with a prompt that explicitly instructs it to generate text in the desired tone and style
  • B. Use a neutralizer to normalize the tone and style of the underlying documents
  • C. Include few-shot examples in the prompt to the LLM
  • D. Fine-tune the LLM on a dataset of desired tone and style


Answer : B

A Generative AI Engineer is creating an LLM-powered application that will need access to up-to-date news articles and stock prices.
The design requires the use of stock prices which are stored in Delta tables and finding the latest relevant news articles by searching the internet.
How should the Generative AI Engineer architect their LLM system?

  • A. Use an LLM to summarize the latest news articles and lookup stock tickers from the summaries to find stock prices.
  • B. Query the Delta table for volatile stock prices and use an LLM to generate a search query to investigate potential causes of the stock volatility.
  • C. Download and store news articles and stock price information in a vector store. Use a RAG architecture to retrieve and generate at runtime.
  • D. Create an agent with tools for SQL querying of Delta tables and web searching, provide retrieved values to an LLM for generation of response.


Answer : D

A Generative AI Engineer is designing a chatbot for a gaming company that aims to engage users on its platform while its users play online video games.
Which metric would help them increase user engagement and retention for their platform?

  • A. Randomness
  • B. Diversity of responses
  • C. Lack of relevance
  • D. Repetition of responses


Answer : B

A company has a typical RAG-enabled, customer-facing chatbot on its website.

Select the correct sequence of components a user's questions will go through before the final output is returned. Use the diagram above for reference.

  • A. 1.embedding model, 2.vector search, 3.context-augmented prompt, 4.response-generating LLM
  • B. 1.context-augmented prompt, 2.vector search, 3.embedding model, 4.response-generating LLM
  • C. 1.response-generating LLM, 2.vector search, 3.context-augmented prompt, 4.embedding model
  • D. 1.response-generating LLM, 2.context-augmented prompt, 3.vector search, 4.embedding model


Answer : A

A team wants to serve a code generation model as an assistant for their software developers. It should support multiple programming languages. Quality is the primary objective.
Which of the Databricks Foundation Model APIs, or models available in the Marketplace, would be the best fit?

  • A. Llama2-70b
  • B. BGE-large
  • C. MPT-7b
  • D. CodeLlama-34B


Answer : D

A Generative AI Engineer is building a RAG application that will rely on context retrieved from source documents that are currently in PDF format. These PDFs can contain both text and images. They want to develop a solution using the least amount of lines of code.
Which Python package should be used to extract the text from the source documents?

  • A. flask
  • B. beautifulsoup
  • C. unstructured
  • D. numpy


Answer : C

A Generative AI Engineer received the following business requirements for an external chatbot.
The chatbot needs to know what types of questions the user asks and routes to appropriate models to answer the questions. For example, the user might ask about upcoming event details. Another user might ask about purchasing tickets for a particular event.
What is an ideal workflow for such a chatbot?

  • A. The chatbot should only look at previous event information
  • B. There should be two different chatbots handling different types of user queries.
  • C. The chatbot should be implemented as a multi-step LLM workflow. First, identify the type of question asked, then route the question to the appropriate model. If it’s an upcoming event question, send the query to a text-to-SQL model. If it’s about ticket purchasing, the customer should be redirected to a payment platform.
  • D. The chatbot should only process payments


Answer : C

A Generative Al Engineer is tasked with developing an application that is based on an open source large language model (LLM). They need a foundation LLM with a large context window.
Which model fits this need?

  • A. DistilBERT
  • B. MPT-30B
  • C. Llama2-70B
  • D. DBRX


Answer : D

Page:    1 / 6   
Exam contains 80 questions

Talk to us!


Have any questions or issues ? Please dont hesitate to contact us

Certlibrary.com is owned by MBS Tech Limited: Room 1905 Nam Wo Hong Building, 148 Wing Lok Street, Sheung Wan, Hong Kong. Company registration number: 2310926
Certlibrary doesn't offer Real Microsoft Exam Questions. Certlibrary Materials do not contain actual questions and answers from Cisco's Certification Exams.
CFA Institute does not endorse, promote or warrant the accuracy or quality of Certlibrary. CFA® and Chartered Financial Analyst® are registered trademarks owned by CFA Institute.
Terms & Conditions | Privacy Policy